# AWQ quantization
Deepseek R1 0528 AWQ
MIT
AWQ-quantized version of DeepSeek R1 0528, supports full-context-length operation on 8x80GB GPUs using vLLM.
Large Language Model
Transformers Supports Multiple Languages

D
cognitivecomputations
145
7
Llama 3.3 70B Instruct AWQ
The AWQ 4-bit quantization version of Llama-3.3-70B-Instruct, developed by Meta AI, supports multiple languages and is suitable for text generation tasks.
Large Language Model
Transformers Supports Multiple Languages

L
kosbu
379.28k
6
Dolphin 2.9 Llama3 70b Awq
AWQ quantized version of Dolphin 2.9 Llama3 70B, suitable for vllm and other inference engines.
Large Language Model
Transformers

D
julep-ai
19
5
Featured Recommended AI Models